A Neurocontrol Paradigm for Intelligent Process Control using Evolutionary Reinforcement Learning

نویسندگان

  • Alex van Eck Conradie
  • Alex van Eck
چکیده

ii DECLARATION I, the undersigned, hereby declare that the work contained in this thesis is my own original work and has not previously in its entirety or in part been submitted at any university for a degree. iii Now to Him who is able to keep us from stumbling, And to bring us faultless Before the presence of His glory with exceeding joy, To God our Saviour, Who alone is wise, Be glory and majesty, Dominion and power, Both now and forever. Amen. iv ACKNOWLEDGEMENTS My Father in heaven for refusing to let me go. Professor Risto Miikkulainen, for his immeasurable guidance at the University of Texas at Austin. Friends and family, for their love, kindness and patience. Loving thanks to my wife, Anje, who put up with "Alex world" for long enough. Balancing multiple business and operational objectives within a comprehensive control strategy is a complex configuration task. Non-linearities and complex multiple process interactions combine as formidable cause-effect interrelationships. A clear understanding of these relationships is often instrumental to meeting the process control objectives. However, such control system configurations are generally conceived in a qualitative manner and with pronounced reliance on past effective configurations (Foss, 1973). Thirty years after Foss' critique, control system configuration remains a largely heuristic affair. Biological methods of processing information are fundamentally different from the methods used in conventional control techniques. Biological neural mechanisms (i.e., intelligent systems) are based on partial models, largely devoid of the system's underlying natural laws. Neural control strategies are carried out without a pure mathematical formulation of the task or the environment. Rather, biological systems rely on knowledge of cause-effect interactions, creating robust control strategies from ill-defined dynamic systems. Dynamic modelling may be either phenomenological or empirical. Phenomenological models are derived from first principles and typically consist of algebraic and differential equations. First principles modelling is both time consuming and expensive. Vast data warehouses of historical plant data make empirical modelling attractive. Singular spectrum analysis (SSA) is a rapid model development technique for identifying dominant state variables from historical plant time series data. Since time series data invariably covers a limited region of the state space, SSA models are almost necessarily partial models. Interpreting and learning causal relationships from dynamic models requires sufficient feedback of the environment's state. Systemisation of the learning task is imperative. Reinforcement learning is a computational approach to understanding and automating goal-directed learning. This thesis aimed …

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Hierarchical Functional Concepts for Knowledge Transfer among Reinforcement Learning Agents

This article introduces the notions of functional space and concept as a way of knowledge representation and abstraction for Reinforcement Learning agents. These definitions are used as a tool of knowledge transfer among agents. The agents are assumed to be heterogeneous; they have different state spaces but share a same dynamic, reward and action space. In other words, the agents are assumed t...

متن کامل

A new Evolutionary Reinforcement Scheme for Stochastic Learning Automata

A stochastic automaton can perform a finite number of actions in a random environment. When a specific action is performed, the environment responds by producing an environment output that is stochastically related to the action. The aim is to design an automaton, using an evolutionary reinforcement scheme (the basis of the learning process), that can determine the best action guided by past ac...

متن کامل

Mini/Micro-Grid Adaptive Voltage and Frequency Stability Enhancement Using Q-learning Mechanism

This paper develops an adaptive control method for controlling frequency and voltage of an islanded mini/micro grid (M/µG) using reinforcement learning method. Reinforcement learning (RL) is one of the branches of the machine learning, which is the main solution method of Markov decision process (MDPs). Among the several solution methods of RL, the Q-learning method is used for solving RL in th...

متن کامل

Using BELBIC based optimal controller for omni-directional threewheel robots model identified by LOLIMOT

In this paper, an intelligent controller is applied to control omni-directional robots motion. First, the dynamics of the three wheel robots, as a nonlinear plant with considerable uncertainties, is identified using an efficient algorithm of training, named LoLiMoT. Then, an intelligent controller based on brain emotional learning algorithm is applied to the identified model. This emotional l...

متن کامل

Intelligent Embodied Agents within a Physically Accurate Environment

EVOLVING INTELLIGENT EMBODIED AGENTS WITHIN A PHYSICALLY ACCURATE ENVIRONMENT By Gene D. Ruebsamen December 2002 This thesis explores the application of evolutionary reinforcement learning techniques for evolving behaviorisms in embodied agents existing within a realistic virtual environment that are subject of the constraints as defined by the Newtonian model of physics. Evolutionary reinforce...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2008